Goto

Collaborating Authors

 prompt text




Automatic Detection of Inauthentic Templated Responses in English Language Assessments

Samant, Yashad, Becker, Lee, Hellman, Scott, Behan, Bradley, Hughes, Sarah, Southerland, Joshua

arXiv.org Artificial Intelligence

Pearson Education, Inc. Author Note Correspondence concerning this article should be addressed to Lee Becker. Pearson affiliated authors can be reached at .@pearson.com. Sarah Hughes can be reached at sarah.hughes1@pearson.com. Joshua Southerland can be reached at josh.southerland@pearson.com In this study, we introduce the automated detection of inauthentic, templated responses (AuDITR) task, describe a machine learning-based approach to this task and illustrate the importance of regularly updating these models in production. Introduction English language proficiency (ELP) tests carry exceptionally high stakes because of how they influence access to employment, education and national residency status.


Attacks and Defenses Against LLM Fingerprinting

Kurian, Kevin, Holland, Ethan, Oesch, Sean

arXiv.org Artificial Intelligence

--As large language models are increasingly deployed in sensitive environments, fingerprinting attacks pose significant privacy and security risks. We present a study of LLM fingerprinting from both offensive and defensive perspectives. Our attack methodology uses reinforcement learning to automatically optimize query selection, achieving better fingerprinting accuracy with only 3 queries compared to randomly selecting 3 queries from the same pool. Our defensive approach employs semantic-preserving output filtering through a secondary LLM to obfuscate model identity while maintaining semantic integrity. The defensive method reduces fingerprinting accuracy across tested models while preserving output quality. These contributions show the potential to improve fingerprinting tools capabilities while providing practical mitigation strategies against fingerprinting attacks. Large language models (LLMs) have become ubiquitous across industries, from customer service chatbots to code generation tools and content creation platforms. As organizations increasingly rely on these models for sensitive applications, the ability to identify which specific model generated a given text--known as LLM fingerprinting--has emerged as a critical security concern.


Automatic Feedback Generation for Short Answer Questions using Answer Diagnostic Graphs

Furuhashi, Momoka, Funayama, Hiroaki, Iwase, Yuya, Matsubayashi, Yuichiroh, Isobe, Yoriko, Nagahama, Toru, Sugawara, Saku, Inui, Kentaro

arXiv.org Artificial Intelligence

Short-reading comprehension questions help students understand text structure but lack effective feedback. Students struggle to identify and correct errors, while manual feedback creation is labor-intensive. This highlights the need for automated feedback linking responses to a scoring rubric for deeper comprehension. Despite advances in Natural Language Processing (NLP), research has focused on automatic grading, with limited work on feedback generation. To address this, we propose a system that generates feedback for student responses. Our contributions are twofold. First, we introduce the first system for feedback on short-answer reading comprehension. These answers are derived from the text, requiring structural understanding. We propose an "answer diagnosis graph," integrating the text's logical structure with feedback templates. Using this graph and NLP techniques, we estimate students' comprehension and generate targeted feedback. Second, we evaluate our feedback through an experiment with Japanese high school students (n=39). They answered two 70-80 word questions and were divided into two groups with minimal academic differences. One received a model answer, the other system-generated feedback. Both re-answered the questions, and we compared score changes. A questionnaire assessed perceptions and motivation. Results showed no significant score improvement between groups, but system-generated feedback helped students identify errors and key points in the text. It also significantly increased motivation. However, further refinement is needed to enhance text structure understanding.


Pronunciation Assessment with Multi-modal Large Language Models

Fu, Kaiqi, Peng, Linkai, Yang, Nan, Zhou, Shuran

arXiv.org Artificial Intelligence

Large language models (LLMs), renowned for their powerful conversational abilities, are widely recognized as exceptional tools in the field of education, particularly in the context of automated intelligent instruction systems for language learning. In this paper, we propose a scoring system based on LLMs, motivated by their positive impact on text-related scoring tasks. Specifically, the speech encoder first maps the learner's speech into contextual features. The adapter layer then transforms these features to align with the text embedding in latent space. The assessment task-specific prefix and prompt text are embedded and concatenated with the features generated by the modality adapter layer, enabling the LLMs to predict accuracy and fluency scores. Our experiments demonstrate that the proposed scoring systems achieve competitive results compared to the baselines on the Speechocean762 datasets. Moreover, we also conducted an ablation study to better understand the contributions of the prompt text and training strategy in the proposed scoring system.


CLIP Model for Images to Textual Prompts Based on Top-k Neighbors

Zhang, Xin, Zhang, Xin, Cai, YeMing, Jia, Tianzhi

arXiv.org Artificial Intelligence

Text-to-image synthesis, a subfield of multimodal generation, has gained significant We propose a cost-effective approach for image-toprompt attention in recent years. We propose a costeffective generation that leverages generative models approach for image-to-prompt generation to generate textual prompts without the need for that leverages generative models to generate textual large amounts of annotated data. Our method allows prompts without the need for large amounts of for direct utilization of the generated prompts or annotated data. We divide our method into two serves as valuable initialization for data-efficient stages: online stage and offline stage. We use a fine-tuning processes. This approach significantly combination of the CLIP model and K-nearest reduces data costs and time consumption while neighbors (KNN) algorithm. The proposed system achieving high quality and diversity in the consists of two main parts: an offline task and an generation of prompts related to input images.


Prompt Engineering-assisted Malware Dynamic Analysis Using GPT-4

Yan, Pei, Tan, Shunquan, Wang, Miaohui, Huang, Jiwu

arXiv.org Artificial Intelligence

Dynamic analysis methods effectively identify shelled, wrapped, or obfuscated malware, thereby preventing them from invading computers. As a significant representation of dynamic malware behavior, the API (Application Programming Interface) sequence, comprised of consecutive API calls, has progressively become the dominant feature of dynamic analysis methods. Though there have been numerous deep learning models for malware detection based on API sequences, the quality of API call representations produced by those models is limited. These models cannot generate representations for unknown API calls, which weakens both the detection performance and the generalization. Further, the concept drift phenomenon of API calls is prominent. To tackle these issues, we introduce a prompt engineering-assisted malware dynamic analysis using GPT-4. In this method, GPT-4 is employed to create explanatory text for each API call within the API sequence. Afterward, the pre-trained language model BERT is used to obtain the representation of the text, from which we derive the representation of the API sequence. Theoretically, this proposed method is capable of generating representations for all API calls, excluding the necessity for dataset training during the generation process. Utilizing the representation, a CNN-based detection model is designed to extract the feature. We adopt five benchmark datasets to validate the performance of the proposed model. The experimental results reveal that the proposed detection algorithm performs better than the state-of-the-art method (TextCNN). Specifically, in cross-database experiments and few-shot learning experiments, the proposed model achieves excellent detection performance and almost a 100% recall rate for malware, verifying its superior generalization performance. The code is available at: github.com/yan-scnu/Prompted_Dynamic_Detection.


Extensible Prompts for Language Models on Zero-shot Language Style Customization

Ge, Tao, Hu, Jing, Dong, Li, Mao, Shaoguang, Xia, Yan, Wang, Xun, Chen, Si-Qing, Wei, Furu

arXiv.org Artificial Intelligence

We propose eXtensible Prompt (X-Prompt) for prompting a large language model (LLM) beyond natural language (NL). X-Prompt instructs an LLM with not only NL but also an extensible vocabulary of imaginary words. Registering new imaginary words allows us to instruct the LLM to comprehend concepts that are difficult to describe with NL words, thereby making a prompt more descriptive. Also, these imaginary words are designed to be out-of-distribution (OOD) robust so that they can be (re)used like NL words in various prompts, distinguishing X-Prompt from soft prompt that is for fitting in-distribution data. We propose context-augmented learning (CAL) to learn imaginary words for general usability, enabling them to work properly in OOD (unseen) prompts. We experiment X-Prompt for zero-shot language style customization as a case study. The promising results of X-Prompt demonstrate its potential to facilitate advanced interaction beyond the natural language interface, bridging the communication gap between humans and LLMs.


Self-Detoxifying Language Models via Toxification Reversal

Leong, Chak Tou, Cheng, Yi, Wang, Jiashuo, Wang, Jian, Li, Wenjie

arXiv.org Artificial Intelligence

Language model detoxification aims to minimize the risk of generating offensive or harmful content in pretrained language models (PLMs) for safer deployment. Existing methods can be roughly categorized as finetuning-based and decoding-based. However, the former is often resource-intensive, while the latter relies on additional components and potentially compromises the generation fluency. In this paper, we propose a more lightweight approach that enables the PLM itself to achieve "self-detoxification". Our method is built upon the observation that prepending a negative steering prompt can effectively induce PLMs to generate toxic content. At the same time, we are inspired by the recent research in the interpretability field, which formulates the evolving contextualized representations within the PLM as an information stream facilitated by the attention layers. Drawing on this idea, we devise a method to identify the toxification direction from the normal generation process to the one prompted with the negative prefix, and then steer the generation to the reversed direction by manipulating the information movement within the attention layers. Experimental results show that our approach, without any fine-tuning or extra components, can achieve comparable performance with state-of-the-art methods.